Fast Computation of the Bezout and Dixon Resultant Matrices

نویسندگان
چکیده

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Fast Computation of the Bezout and Dixon Resultant Matrices

Efficient algorithms are derived for computing the entries of the Bezout resultant matrix for two univariate polynomials of degree n and for calculating the entries of the Dixon–Cayley resultant matrix for three bivariate polynomials of bidegree (m, n). Standard methods based on explicit formulas require O(n3) additions and multiplications to compute all the entries of the Bezout resultant matr...

متن کامل

Constructing Sylvester-type resultant matrices using the Dixon formulation

A new method for constructing Sylvester-type resultant matrices for multivariate elimination is proposed. Unlike sparse resultant constructions discussed recently in the literature or the Macaulay resultant construction, the proposed method does not explicitly use the support of a polynomial system in the construction. Instead, a multiplier set for each polynomial is obtained from the Dixon for...

متن کامل

Cayley-Dixon Resultant Matrices of Multi-univariate Composed Polynomials

The behavior of the Cayley-Dixon resultant construction and the structure of Dixon matrices are analyzed for composed polynomial systems constructed from a multivariate system in which each variable is substituted by a univariate polynomial in a distinct variable. It is shown that a Dixon projection operator (a multiple of the resultant) of the composed system can be expressed as a power of the...

متن کامل

Computation of the q-th roots of circulant matrices

In this paper, we investigate the reduced form of circulant matrices and we show that the problem of computing the q-th roots of a nonsingular circulant matrix A can be reduced to that of computing the q-th roots of two half size matrices B - C and B + C.

متن کامل

Fast Computation of Moore-Penrose Inverse Matrices

Many neural learning algorithms require to solve large least square systems in order to obtain synaptic weights. Moore-Penrose inverse matrices allow for solving such systems, even with rank deficiency, and they provide minimum-norm vectors of synaptic weights, which contribute to the regularization of the input-output mapping. It is thus of interest to develop fast and accurate algorithms for ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Journal of Symbolic Computation

سال: 2002

ISSN: 0747-7171

DOI: 10.1006/jsco.2001.0462